global question
- Europe > Spain > Basque Country > Álava Province > Vitoria-Gasteiz (0.04)
- South America > Brazil (0.04)
- South America > Argentina (0.04)
- (11 more...)
- Leisure & Entertainment (1.00)
- Education (1.00)
- Media > Film (0.67)
- Europe > Spain > Basque Country > Álava Province > Vitoria-Gasteiz (0.04)
- South America > Brazil (0.04)
- South America > Argentina (0.04)
- (11 more...)
- Leisure & Entertainment (1.00)
- Education (1.00)
- Media > Film (0.67)
BertaQA: How Much Do Language Models Know About Local Culture?
Etxaniz, Julen, Azkune, Gorka, Soroa, Aitor, de Lacalle, Oier Lopez, Artetxe, Mikel
Large Language Models (LLMs) exhibit extensive knowledge about the world, but most evaluations have been limited to global or anglocentric subjects. This raises the question of how well these models perform on topics relevant to other cultures, whose presence on the web is not that prominent. To address this gap, we introduce BertaQA, a multiple-choice trivia dataset that is parallel in English and Basque. The dataset consists of a local subset with questions pertinent to the Basque culture, and a global subset with questions of broader interest. We find that state-of-the-art LLMs struggle with local cultural knowledge, even as they excel on global topics. However, we show that continued pre-training in Basque significantly improves the models' performance on Basque culture, even when queried in English. To our knowledge, this is the first solid evidence of knowledge transfer from a low-resource to a high-resource language. Our analysis sheds light on the complex interplay between language and knowledge, and reveals that some prior findings do not fully hold when reassessed on local topics. Our dataset and evaluation code are available under open licenses at https://github.com/juletx/BertaQA.
- Europe > Spain > Basque Country (0.04)
- South America > Brazil (0.04)
- South America > Argentina (0.04)
- (10 more...)
- Leisure & Entertainment (0.67)
- Government (0.67)
- Education (0.66)
Deep Neural Networks Evolve Human-like Attention Distribution during Reading Comprehension
Attention is a key mechanism for information selection in both biological brains and many state-of-the-art deep neural networks (DNNs). Here, we investigate whether humans and DNNs allocate attention in comparable ways when reading a text passage to subsequently answer a specific question. We analyze 3 transformer-based DNNs that reach human-level performance when trained to perform the reading comprehension task. We find that the DNN attention distribution quantitatively resembles human attention distribution measured by fixation times. Human readers fixate longer on words that are more relevant to the question-answering task, demonstrating that attention is modulated by top-down reading goals, on top of lower-level visual and text features of the stimulus. Further analyses reveal that the attention weights in DNNs are also influenced by both top-down reading goals and lower-level stimulus features, with the shallow layers more strongly influenced by lower-level text features and the deep layers attending more to task-relevant words. Additionally, deep layers' attention to task-relevant words gradually emerges when pre-trained DNN models are fine-tuned to perform the reading comprehension task, which coincides with the improvement in task performance. These results demonstrate that DNNs can evolve human-like attention distribution through task optimization, which suggests that human attention during goal-directed reading comprehension is a consequence of task optimization.
- Asia > China > Zhejiang Province > Hangzhou (0.04)
- North America > United States (0.04)
- Europe > Switzerland > Basel-City > Basel (0.04)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (1.00)
- Education > Assessment & Standards > Student Performance (1.00)
- Education > Educational Setting (0.93)
- Health & Medicine > Therapeutic Area (0.68)